Optimization Based Layer-Wise Pruning Threshold Method for Accelerating Convolutional Neural Networks

نویسندگان

چکیده

Among various network compression methods, pruning has developed rapidly due to its superior performance. However, the trivial threshold limits performance of pruning. Most conventional methods are based on well-known hard or soft techniques that rely time-consuming handcrafted tests domain experience. To mitigate these issues, we propose a simple yet effective general method from an optimization point view. Specifically, problem is formulated as constrained program minimizes size each layer. More importantly, our together with works achieves better across scenarios many advanced benchmarks. Notably, for L1-norm algorithm VGG-16, higher FLOPs reductions without utilizing sensibility analysis. The ratio boosts 34% 53%, which huge improvement. Similar experiments ResNet-56 reveal that, even compact networks, competitive skipping any sensitive layers.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Pruning Convolutional Neural Networks for Resource Efficient Transfer Learning

We propose a new framework for pruning convolutional kernels in neural networks to enable efficient inference, focusing on transfer learning where large and potentially unwieldy pretrained networks are adapted to specialized tasks. We interleave greedy criteria-based pruning with fine-tuning by backpropagation—a computationally efficient procedure that maintains good generalization in the prune...

متن کامل

Pruning Convolutional Neural Networks for Image Instance Retrieval

In this work, we focus on the problem of image instance retrieval with deep descriptors extracted from pruned Convolutional Neural Networks (CNN). The objective is to heavily prune convolutional edges while maintaining retrieval performance. To this end, we introduce both data-independent and data-dependent heuristics to prune convolutional edges, and evaluate their performance across various c...

متن کامل

Pruning Convolutional Neural Networks for Resource Efficient Inference

We propose a new formulation for pruning convolutional kernels in neural networks to enable efficient inference. We interleave greedy criteria-based pruning with finetuning by backpropagation—a computationally efficient procedure that maintains good generalization in the pruned network. We propose a new criterion based on Taylor expansion that approximates the change in the cost function induce...

متن کامل

Net-Trim: A Layer-wise Convex Pruning of Deep Neural Networks

and quantum settings Model reduction is a highly desirable process for deep neural networks. While large networks are theoretically capable of learning arbitrarily complex models, overfitting and model redundancy negatively affects the prediction accuracy and model variance. Net-Trim is a layer-wise convex framework to prune (sparsify) deep neural networks. The method is applicable to neural ne...

متن کامل

Introducing a method for extracting features from facial images based on applying transformations to features obtained from convolutional neural networks

In pattern recognition, features are denoting some measurable characteristics of an observed phenomenon and feature extraction is the procedure of measuring these characteristics. A set of features can be expressed by a feature vector which is used as the input data of a system. An efficient feature extraction method can improve the performance of a machine learning system such as face recognit...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Mathematics

سال: 2023

ISSN: ['2227-7390']

DOI: https://doi.org/10.3390/math11153311